Goto

Collaborating Authors

 Cape Breton County


QiVC-Net: Quantum-Inspired Variational Convolutional Network, with Application to Biosignal Classification

Golnari, Amin, Yousefi, Jamileh, Moheimani, Reza, Sanei, Saeid

arXiv.org Artificial Intelligence

This work introduces the quantum-inspired variational convolution (QiVC) framework, a novel learning paradigm that integrates principles of probabilistic inference, variational optimization, and quantum-inspired transformations within convolutional architectures. The central innovation of QiVC lies in its quantum-inspired rotated ensemble (QiRE) mechanism. QiRE performs differentiable low-dimensional subspace rotations of convolutional weights, analogously to quantum state evolution. This approach enables structured uncertainty modeling while preserving the intrinsic geometry of the parameter space, resulting in more expressive, stable, and uncertainty-aware representations. To demonstrate its practical potential, the concept is instantiated in a QiVC-based convolutional network (QiVC-Net) and evaluated in the context of biosignal classification, focusing on phonocardiogram (PCG) recordings, a challenging domain characterized by high noise, inter-subject variability, and often imbalanced data. The proposed QiVC-Net integrates an architecture in which the QiVC layer does not introduce additional parameters, instead performing an ensemble rotation of the convolutional weights through a structured mechanism ensuring robustness without added highly computational burden. Experiments on two benchmark datasets, PhysioNet CinC 2016 and PhysioNet CirCor DigiScope 2022, show that QiVC-Net achieves state-of-the-art performance, reaching accuracies of 97.84% and 97.89%, respectively. These findings highlight the versatility of the QiVC framework and its promise for advancing uncertainty-aware modeling in real-world biomedical signal analysis. The implementation of the QiVConv layer is openly available in GitHub.


Charting the Future of Scholarly Knowledge with AI: A Community Perspective

Jiomekong, Azanzi, McGinty, Hande Küçük, Mills, Keith G., Oelen, Allard, Rajabi, Enayat, McElroy, Harry, Christou, Antrea, Saini, Anmol, Zebaze, Janice Anta, Kim, Hannah, Jacyszyn, Anna M., Auer, Sören

arXiv.org Artificial Intelligence

Scholarly work and communication encompass the entire system in which research and creative works are created, evaluated for quality, disseminated to the academic community and beyond, used, and preserved for future use. It includes formal publications, such as journal articles and books, as well as informal sharing through preprints, conference presentations, data sharing, and broader engagement with scholarly works and research outputs. Scholarly knowledge serves as the primary engine of progress, shaping our world and guiding our collective future. It forms the backbone of technological advancement, public health systems, and sustainable environmental practices. Obtained through rigorous methods of observation, experimentation, and validation, it is a reliable resource that helps societies solve complex problems and improve the quality of life by achieving sustainable development goals (SDGs) [6]. To be truly useful, scholarly knowledge must first be systematically extracted and organized. However, the scholarly community of today faces the problem of an overload of scientific papers in their respective domains. There is an increasing number of papers published every year (currently, 3 million), in addition to more than 200 million papers that have already been published . This gives rise to the research question: "How can we provide a reliable and living scholarly knowledge base that empowers researchers to query, synthesize, and analyze the vast body of scholarly knowledge?"


ComplexFuncBench: Exploring Multi-Step and Constrained Function Calling under Long-Context Scenario

Zhong, Lucen, Du, Zhengxiao, Zhang, Xiaohan, Hu, Haiyi, Tang, Jie

arXiv.org Artificial Intelligence

Enhancing large language models (LLMs) with real-time APIs can help generate more accurate and up-to-date responses. However, evaluating the function calling abilities of LLMs in real-world scenarios remains under-explored due to the complexity of data collection and evaluation. In this work, we introduce ComplexFuncBench, a benchmark for complex function calling across five real-world scenarios. Compared to existing benchmarks, ComplexFuncBench encompasses multi-step and constrained function calling, which requires long-parameter filing, parameter value reasoning, and 128k long context. Additionally, we propose an automatic framework, ComplexEval, for quantitatively evaluating complex function calling tasks. Through comprehensive experiments, we demonstrate the deficiencies of state-of-the-art LLMs in function calling and suggest future directions for optimizing these capabilities. The data and code are available at \url{https://github.com/THUDM/ComplexFuncBench}.


LLMs in the Imaginarium: Tool Learning through Simulated Trial and Error

Wang, Boshi, Fang, Hao, Eisner, Jason, Van Durme, Benjamin, Su, Yu

arXiv.org Artificial Intelligence

Tools are essential for large language models (LLMs) to acquire up-to-date information and take consequential actions in external environments. Existing work on tool-augmented LLMs primarily focuses on the broad coverage of tools and the flexibility of adding new tools. However, a critical aspect that has surprisingly been understudied is simply how accurately an LLM uses tools for which it has been trained. We find that existing LLMs, including GPT-4 and open-source LLMs specifically fine-tuned for tool use, only reach a correctness rate in the range of 30% to 60%, far from reliable use in practice. We propose a biologically inspired method for tool-augmented LLMs, simulated trial and error (STE), that orchestrates three key mechanisms for successful tool use behaviors in the biological system: trial and error, imagination, and memory. Specifically, STE leverages an LLM's 'imagination' to simulate plausible scenarios for using a tool, after which the LLM interacts with the tool to learn from its execution feedback. Both short-term and long-term memory are employed to improve the depth and breadth of the exploration, respectively. Comprehensive experiments on ToolBench show that STE substantially improves tool learning for LLMs under both in-context learning and fine-tuning settings, bringing a boost of 46.7% to Mistral-Instruct-7B and enabling it to outperform GPT-4. We also show effective continual learning of tools via a simple experience replay strategy.


The Perspective of Software Professionals on Algorithmic Racism

Santos, Ronnie de Souza, de Lima, Luiz Fernando, Magalhaes, Cleyton

arXiv.org Artificial Intelligence

Context. Algorithmic racism is the term used to describe the behavior of technological solutions that constrains users based on their ethnicity. Lately, various data-driven software systems have been reported to discriminate against Black people, either for the use of biased data sets or due to the prejudice propagated by software professionals in their code. As a result, Black people are experiencing disadvantages in accessing technology-based services, such as housing, banking, and law enforcement. Goal. This study aims to explore algorithmic racism from the perspective of software professionals. Method. A survey questionnaire was applied to explore the understanding of software practitioners on algorithmic racism, and data analysis was conducted using descriptive statistics and coding techniques. Results. We obtained answers from a sample of 73 software professionals discussing their understanding and perspectives on algorithmic racism in software development. Our results demonstrate that the effects of algorithmic racism are well-known among practitioners. However, there is no consensus on how the problem can be effectively addressed in software engineering. In this paper, some solutions to the problem are proposed based on the professionals' narratives. Conclusion. Combining technical and social strategies, including training on structural racism for software professionals, is the most promising way to address the algorithmic racism problem and its effects on the software solutions delivered to our society.


Inverse design of nano-photonic wavelength demultiplexer with a deep neural network approach

Yuan, Mengwei, Yang, Gang, Song, Shijie, Zhou, Luping, Minasian, Robert, Yi, Xiaoke

arXiv.org Artificial Intelligence

In this paper, we propose a pre-trained-combined neural network (PTCN) as a comprehensive solution to the inverse design of an integrated photonic circuit. By utilizing both the initially pre-trained inverse and forward model with a joint training process, our PTCN model shows remarkable tolerance to the quantity and quality of the training data. As a proof of concept demonstration, the inverse design of a wavelength demultiplexer is used to verify the effectiveness of the PTCN model. The correlation coefficient of the prediction by the presented PTCN model remains greater than 0.974 even when the size of training data is decreased to 17%. The experimental results show a good agreement with predictions, and demonstrate a wavelength demultiplexer with an ultra-compact footprint, a high transmission efficiency with a transmission loss of -2dB, a low reflection of -10dB, and low crosstalk around -7dB simultaneously.


Strategyproof Peer Selection: Mechanisms, Analyses, and Experiments

Aziz, Haris (Data61 and University of New South Wales) | Lev, Omer (University of Toronto) | Mattei, Nicholas (Data61 and University of New South Wales) | Rosenschein, Jeffrey S. (The Hebrew University of Jerusalem) | Walsh, Toby (Data61 and University of New South Wales)

AAAI Conferences

We study an important crowdsourcing setting where agents evaluate one another and, based on these evaluations, a subset of agents are selected. This setting is ubiquitous when peer review is used for distributing awards in a team, allocating funding to scientists, and selecting publications for conferences. The fundamental challenge when applying crowdsourcing in these settings is that agents may misreport their reviews of others to increase their chances of being selected. We propose a new strategyproof (impartial) mechanism called Dollar Partition that satisfies desirable axiomatic properties. We then show, using a detailed experiment with parameter values derived from target real world domains, that our mechanism performs better on average, and in the worst case, than other strategyproof mechanisms in the literature.


Multiple decision trees

Kwok, Suk Wah, Carter, Chris

arXiv.org Machine Learning

This paper describes experiments, on two domains, to investigate the effect of averaging over predictions of multiple decision trees, instead of using a single tree. Other authors have pointed out theoretical and commonsense reasons for preferring the multiple tree approach. Ideally, we would like to consider predictions from all trees, weighted by their probability. However, there is a vast number of different trees, and it is difficult to estimate the probability of each tree. We sidestep the estimation problem by using a modified version of the ID3 algorithm to build good trees, and average over only these trees. Our results are encouraging. For each domain, we managed to produce a small number of good trees. We find that it is best to average across sets of trees with different structure; this usually gives better performance than any of the constituent trees, including the ID3 tree.


Eliminating the Weakest Link: Making Manipulation Intractable?

Davies, Jessica (University of Toronto) | Narodytska, Nina (NICTA and University of New South Wales) | Walsh, Toby (NICTA and University of New South Wales)

AAAI Conferences

Successive elimination of candidates is often a route to making manipulation intractable to compute. We prove that eliminating candidates does not necessarily increase the computational complexity of manipulation. However, for many voting rules used in practice, the computational complexity increases. For example, it is already known that it is NP-hard to compute how a single voter can manipulate the result of single transferable voting (the elimination version of plurality voting). We show here that it is NP-hard to compute how a single voter can manipulate the result of the elimination version of veto voting, of the closely related Coombs’ rule, and of the elimination versions of a general class of scoring rules.


Propagating Conjunctions of AllDifferent Constraints

Bessiere, Christian (LIRMM, CNRS) | Katsirelos, George (CRIL-CNRS) | Narodytska, Nina (NICTA and UNSW) | Quimper, Claude-Guy (Universite Laval) | Walsh, Toby (NICTA and UNSW)

AAAI Conferences

We study propagation algorithms for the conjunction of two AllDifferent constraints. Solutions of an AllDifferent constraint can be seen as perfect matchings on the variable/value bipartite graph. Therefore, we investigate the problem of finding simultaneous bipartite matchings. We present an extension of the famous Hall theorem which characterizes when simultaneous bipartite matchings exists. Unfortunately, finding such matchings is NP-hard in general. However, we prove a surprising result that finding a simultaneous matching on a convex bipartite graph takes just polynomial time. Based on this theoretical result, we provide the first polynomial time bound consistency algorithm for the conjunction of two AllDifferent constraints. We identify a pathological problem on which this propagator is exponentially faster compared to existing propagators. Our experiments show that this new propagator can offer significant benefits over existing methods.